A Modified Editing k-nearest Neighbor Rule

نویسندگان

  • Ruiqin Chang
  • Zheng Pei
  • Chao Zhang
چکیده

Classification of objects is an important area in a variety of fields and applications. Many different methods are available to make a decision in those cases. The knearest neighbor rule (k-NN) is a well-known nonparametric decision procedure. Classification rules based on the k-NN have already been proposed and applied in diverse substantive areas. The editing k-NN proposed by Wilson would be an important one. In this rule, editing the reference set is first performed, every sample in the reference set is classified by using the k-NN rule and the set is formed by eliminating it from the reference set. All the samples mistakenly classified are then deleted from the reference set. Afterward, any input sample is classified using the k-NN rule and the edited reference set. Obviously, the editing k-nearest neighbors classifier (EK -NN) consists of the k-nearest neighbor classifier and an editing reference set. However, the editing reference set gained by this method is only a subset of the reference set. This may result in the loss of some important information and decline of classification accuracy. In this paper, we focus on modifying the editing reference set of EK -NN, the new editing set in our method consists of subsets of the reference set and testing set, such subsets are received by classifying every sample in the reference set and testing set by using the k-NN rule and removing misclassified samples from reference set and testing set, respectively. Advantages of our method are to reduce the loss of information and improve the recognition rate. Comparisons and analysis of the experimental results demonstrate the capability of the proposed algorithm.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Evidential Editing K-Nearest Neighbor Classifier

One of the difficulties that arises when using the K-nearest neighbor rule is that each of the labeled training samples is given equal importance in deciding the class of the query pattern to be classified, regardless of their typicality. In this paper, the theory of belief functions is introduced into the K-nearest neighbor rule to develop an evidential editing version of this algorithm. An ev...

متن کامل

Asymptotic Properties of Nearest Neighbor Rules Using Edited Data

The convergence properties of a nearest neighbor rule that uses an editing procedure to reduce the number of preclassified samples and to improve the performance of the rule are developed. Editing of the preclassified samples using the three-nearest neighbor rule followed by classification using the single-nearest neighbor rule with the remaining preclassified samples appears to produce a decis...

متن کامل

FUZZY K-NEAREST NEIGHBOR METHOD TO CLASSIFY DATA IN A CLOSED AREA

Clustering of objects is an important area of research and application in variety of fields. In this paper we present a good technique for data clustering and application of this Technique for data clustering in a closed area. We compare this method with K-nearest neighbor and K-means.  

متن کامل

Application of Proximity Graphs to Editing Nearest Neighbor Decision Rules

Non-parametric decision rules, such as the nearest neighbor (NN) rule, are attractive because no a priori knowledge is required concerning the underlying distributions of the data. Two traditional criticisms directed at the NN-rule concern the large amounts of storage and computation involved due to the apparent necessity to store all the sample (training) data. Thus there has been considerable...

متن کامل

Fitness functions in editing k-NN reference set by genetic algorithms

-In a previous paper the use of GAs as an editing technique for the k-nearest neighbor (k-NN) classification technique has been suggested. Here we are looking at different fitness functions. An experimental study with the IRIS data set and with a medical data set has been carried out. Best results (smallest subsets with highest test classification accuracy) have been obtained by including in th...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:
  • JCP

دوره 6  شماره 

صفحات  -

تاریخ انتشار 2011